Strict Monotonicity of Sum of Squares Error and Normalized Cut in the Lattice of Clusterings
نویسنده
چکیده
Sum of Squares Error and Normalized Cut are two widely used clustering functional. It is known their minimum values are monotone with respect to the input number of clusters and this monotonicity does not allow for a simple automatic selection of a correct number of clusters. Here we study monotonicity not just on the minimizers but on the entire clustering lattice. We show the value of Sum of Squares Error is strictly monotone under the strict refinement relation of clusterings and we obtain data-dependent bounds on the difference between the value of a clustering and one of its refinements. Using analogous techniques we show the value of Normalized Cut is strictly anti-monotone. These results imply that even if we restrict our solutions to form a chain of clustering, like the one we get from hierarchical algorithms, we cannot rely on the functional values in order to choose the number of clusters. By using these results we get some data-dependent bounds on the difference of the values of any two clusterings.
منابع مشابه
A comparative performance of gray level image thresholding using normalized graph cut based standard S membership function
In this research paper, we use a normalized graph cut measure as a thresholding principle to separate an object from the background based on the standard S membership function. The implementation of the proposed algorithm known as fuzzy normalized graph cut method. This proposed algorithm compared with the fuzzy entropy method [25], Kittler [11], Rosin [21], Sauvola [23] and Wolf [33] method. M...
متن کاملSum Formula for Maximal Abstract Monotonicity and Abstract Rockafellar’s Surjectivity Theorem
In this paper, we present an example in which the sum of two maximal abstract monotone operators is maximal. Also, we shall show that the necessary condition for Rockafellar’s surjectivity which was obtained in ([19], Theorem 4.3) can be sufficient.
متن کاملAn Evolutionary Multi-objective Discretization based on Normalized Cut
Learning models and related results depend on the quality of the input data. If raw data is not properly cleaned and structured, the results are tending to be incorrect. Therefore, discretization as one of the preprocessing techniques plays an important role in learning processes. The most important challenge in the discretization process is to reduce the number of features’ values. This operat...
متن کاملBEST APPROXIMATION IN QUASI TENSOR PRODUCT SPACE AND DIRECT SUM OF LATTICE NORMED SPACES
We study the theory of best approximation in tensor product and the direct sum of some lattice normed spacesX_{i}. We introduce quasi tensor product space anddiscuss about the relation between tensor product space and thisnew space which we denote it by X boxtimesY. We investigate best approximation in direct sum of lattice normed spaces by elements which are not necessarily downwardor upward a...
متن کاملSome remarks on the sum of the inverse values of the normalized signless Laplacian eigenvalues of graphs
Let G=(V,E), $V={v_1,v_2,ldots,v_n}$, be a simple connected graph with $%n$ vertices, $m$ edges and a sequence of vertex degrees $d_1geqd_2geqcdotsgeq d_n>0$, $d_i=d(v_i)$. Let ${A}=(a_{ij})_{ntimes n}$ and ${%D}=mathrm{diag }(d_1,d_2,ldots , d_n)$ be the adjacency and the diagonaldegree matrix of $G$, respectively. Denote by ${mathcal{L}^+}(G)={D}^{-1/2}(D+A) {D}^{-1/2}$ the normalized signles...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2013